2026-04-10 19:28:56,981 [ 368045 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2026-04-10 19:28:56,981 [ 368045 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:79, check_args_and_update_paths) 2026-04-10 19:28:56,981 [ 368045 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:90, check_args_and_update_paths) 2026-04-10 19:28:56,981 [ 368045 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:92, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_af1sbo --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests test_mask_sensitive_info/test.py::test_create_table test_mask_sensitive_info/test.py::test_table_functions test_reload_clusters_config/test.py::test_delete_cluster test_reload_clusters_config/test.py::test_update_one_cluster -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [4 items] scheduling tests via LoadFileScheduling test_mask_sensitive_info/test.py::test_create_table test_reload_clusters_config/test.py::test_delete_cluster [gw0] [ 25%] FAILED test_mask_sensitive_info/test.py::test_create_table test_mask_sensitive_info/test.py::test_table_functions [gw0] [ 50%] FAILED test_mask_sensitive_info/test.py::test_table_functions [gw1] [ 75%] PASSED test_reload_clusters_config/test.py::test_delete_cluster test_reload_clusters_config/test.py::test_update_one_cluster [gw1] [100%] FAILED test_reload_clusters_config/test.py::test_update_one_cluster =================================== FAILURES =================================== ______________________________ test_create_table _______________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 def test_create_table(): password = new_password() azure_conn_string = cluster.env_variables["AZURITE_CONNECTION_STRING"] account_key_pattern = re.compile("AccountKey=.*?(;|$)") masked_azure_conn_string = re.sub( account_key_pattern, "AccountKey=[HIDDEN]\\1", azure_conn_string ) masked_sas_conn_string = 'BlobEndpoint=https://my-endpoint/;SharedAccessSignature=[HIDDEN]' azure_storage_account_url = cluster.env_variables["AZURITE_STORAGE_ACCOUNT_URL"] azure_account_name = "devstoreaccount1" azure_account_key = "Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==" table_engines = [ f"MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '{password}')", f"PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', '{password}')", f"MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', '{password}')", f"S3('http://minio1:9001/root/data/test1.csv')", f"S3('http://minio1:9001/root/data/test2.csv', 'CSV')", f"S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip')", f"S3('http://minio1:9001/root/data/test4.csv', 'minio', '{password}', 'CSV')", f"S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', '{password}', 'CSV', 'gzip')", f"MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', table = 'mysql_table', user = 'mysql_user', password = '{password}')", f"MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = '{password}', table = 'mysql_table', user = 'mysql_user')", f"MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table')", f"PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', table = 'postgres_table', user = 'postgres_user', password = '{password}')", f"MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = '{password}')", f"S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = '{password}', format = 'CSV')", f"S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip')", f"S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', '{password}')", ( f"DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", "DNS_ERROR", ), f"S3Queue('http://minio1:9001/root/data/', 'CSV') settings mode = 'ordered'", f"S3Queue('http://minio1:9001/root/data/', 'CSV', 'gzip') settings mode = 'ordered'", f"S3Queue('http://minio1:9001/root/data/', 'minio', '{password}', 'CSV') settings mode = 'ordered'", f"S3Queue('http://minio1:9001/root/data/', 'minio', '{password}', 'CSV', 'gzip') settings mode = 'ordered'", ( f"Iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", "DNS_ERROR", ), ( f"IcebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", "DNS_ERROR", ), f"AzureBlobStorage('{azure_conn_string}', 'cont', 'test_simple.csv', 'CSV')", f"AzureBlobStorage('{azure_conn_string}', 'cont', 'test_simple_1.csv', 'CSV', 'none')", f"AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_2.csv', '{azure_account_name}', '{azure_account_key}')", f"AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_3.csv', '{azure_account_name}', '{azure_account_key}', 'CSV')", f"AzureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_4.csv', '{azure_account_name}', '{azure_account_key}', 'CSV', 'none')", f"AzureQueue('{azure_conn_string}', 'cont', '*', 'CSV') SETTINGS mode = 'unordered'", f"AzureQueue('{azure_conn_string}', 'cont', '*', 'CSV', 'none') SETTINGS mode = 'unordered'", f"AzureQueue('{azure_storage_account_url}', 'cont', '*', '{azure_account_name}', '{azure_account_key}', 'CSV') SETTINGS mode = 'unordered'", f"AzureQueue('{azure_storage_account_url}', 'cont', '*', '{azure_account_name}', '{azure_account_key}', 'CSV', 'none') SETTINGS mode = 'unordered'", ( f"AzureBlobStorage('BlobEndpoint=https://my-endpoint/;SharedAccessSignature=sp=r&st=2025-09-29T14:58:11Z&se=2025-09-29T00:00:00Z&spr=https&sv=2022-11-02&sr=c&sig=SECRET%SECRET%SECRET%SECRET', 'exampledatasets', 'example.csv')", "STD_EXCEPTION", ), f"Kafka() SETTINGS kafka_broker_list = '127.0.0.1', kafka_topic_list = 'topic', kafka_group_name = 'group', kafka_format = 'JSONEachRow', format_avro_schema_registry_url = 'http://schema_user:{password}@'", f"Kafka() SETTINGS kafka_broker_list = '127.0.0.1', kafka_topic_list = 'topic', kafka_group_name = 'group', kafka_format = 'JSONEachRow', format_avro_schema_registry_url = 'http://schema_user:{password}@domain.com'", ] def make_test_case(i): table_name = f"table{i}" table_engine = table_engines[i] error = None if isinstance(table_engine, tuple): table_engine, error = table_engine query = f"CREATE TABLE {table_name} (x int) ENGINE = {table_engine}" return table_name, query, error # Generate test cases as a list of tuples (table_name, query, error). test_cases = [make_test_case(i) for i in range(len(table_engines))] for table_name, query, error in test_cases: if error: > assert error in node.query_and_get_error(query) E assert 'DNS_ERROR' in "Received exception from server (version 25.3.8):\nCode: 56. DB::Exception: Received from 172.16.2.6:9000. DB::Exception: Unknown table engine DeltaLake. Stack trace:\n\n0. DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000d7ed77b\n1. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000008292ccc\n2. DB::Exception::Exception(int, FormatStringHelperImpl::type>, String&) @ 0x00000000082a8d6b\n3. DB::StorageFactory::get(DB::ASTCreateQuery const&, String const&, std::shared_ptr, std::shared_ptr, DB::ColumnsDescription const&, DB::ConstraintsDescription const&, DB::LoadingStrictnessLevel, bool) const @ 0x00000000122c85cf\n4. DB::InterpreterCreateQuery::doCreateTable(DB::ASTCreateQuery&, DB::InterpreterCreateQuery::TableProperties const&, std::unique_ptr>&, DB::LoadingStrictnessLevel) @ 0x000000001187aa00\n5. DB::InterpreterCreateQuery::createTable(DB::ASTCreateQuery&) @ 0x000000001186f6c2\n6. DB::InterpreterCreateQuery::execute() @ 0x0000000011882c18\n7. DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*, std::shared_ptr&) @ 0x0000000011cadfab\n8. DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000011ca89c4\n9. DB::TCPHandler::runImpl() @ 0x0000000012fed5ac\n10. DB::TCPHandler::run() @ 0x000000001300b799\n11. Poco::Net::TCPServerConnection::start() @ 0x0000000016744027\n12. Poco::Net::TCPServerDispatcher::run() @ 0x0000000016744479\n13. Poco::PooledThread::run() @ 0x00000000167107db\n14. Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001670ecbd\n15. ? @ 0x00007ff928f97ac3\n16. ? @ 0x00007ff929029850\n. (UNKNOWN_STORAGE)\n(query: CREATE TABLE table16 (x int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'EI9HJYOR677QVQBS'))\n" E + where "Received exception from server (version 25.3.8):\nCode: 56. DB::Exception: Received from 172.16.2.6:9000. DB::Exception: Unknown table engine DeltaLake. Stack trace:\n\n0. DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000d7ed77b\n1. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000008292ccc\n2. DB::Exception::Exception(int, FormatStringHelperImpl::type>, String&) @ 0x00000000082a8d6b\n3. DB::StorageFactory::get(DB::ASTCreateQuery const&, String const&, std::shared_ptr, std::shared_ptr, DB::ColumnsDescription const&, DB::ConstraintsDescription const&, DB::LoadingStrictnessLevel, bool) const @ 0x00000000122c85cf\n4. DB::InterpreterCreateQuery::doCreateTable(DB::ASTCreateQuery&, DB::InterpreterCreateQuery::TableProperties const&, std::unique_ptr>&, DB::LoadingStrictnessLevel) @ 0x000000001187aa00\n5. DB::InterpreterCreateQuery::createTable(DB::ASTCreateQuery&) @ 0x000000001186f6c2\n6. DB::InterpreterCreateQuery::execute() @ 0x0000000011882c18\n7. DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*, std::shared_ptr&) @ 0x0000000011cadfab\n8. DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000011ca89c4\n9. DB::TCPHandler::runImpl() @ 0x0000000012fed5ac\n10. DB::TCPHandler::run() @ 0x000000001300b799\n11. Poco::Net::TCPServerConnection::start() @ 0x0000000016744027\n12. Poco::Net::TCPServerDispatcher::run() @ 0x0000000016744479\n13. Poco::PooledThread::run() @ 0x00000000167107db\n14. Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001670ecbd\n15. ? @ 0x00007ff928f97ac3\n16. ? @ 0x00007ff929029850\n. (UNKNOWN_STORAGE)\n(query: CREATE TABLE table16 (x int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'EI9HJYOR677QVQBS'))\n" = >("CREATE TABLE table16 (x int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'EI9HJYOR677QVQBS')") E + where > = .query_and_get_error test_mask_sensitive_info/test.py:309: AssertionError ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ------------------------------ Captured log setup ------------------------------ 2026-04-10 19:29:03.853000 [ 635 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:121, run_and_check) 2026-04-10 19:29:03.884000 [ 635 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 19:29:03.885000 [ 635 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2026-04-10 19:29:03.885000 [ 635 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2026-04-10 19:29:03.885000 [ 635 ] DEBUG : Command:[docker network prune --force] (cluster.py:121, run_and_check) 2026-04-10 19:29:03.911000 [ 635 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:121, run_and_check) 2026-04-10 19:29:03.914000 [ 635 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:145, run_and_check) 2026-04-10 19:29:03.915000 [ 635 ] INFO : Running tests in /ClickHouse/tests/integration/test_mask_sensitive_info/test.py (cluster.py:2748, start) 2026-04-10 19:29:03.916000 [ 635 ] DEBUG : Cluster start called. is_up=False (cluster.py:2755, start) 2026-04-10 19:29:03.940000 [ 635 ] DEBUG : Docker networks for project roottestmasksensitiveinfo-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 19:29:03.968000 [ 635 ] DEBUG : Docker containers for project roottestmasksensitiveinfo-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 19:29:03.996000 [ 635 ] DEBUG : Docker volumes for project roottestmasksensitiveinfo-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 19:29:03.997000 [ 635 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 19:29:04.025000 [ 635 ] DEBUG : Docker networks for project roottestmasksensitiveinfo-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 19:29:04.053000 [ 635 ] DEBUG : Docker containers for project roottestmasksensitiveinfo-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 19:29:04.081000 [ 635 ] DEBUG : Docker volumes for project roottestmasksensitiveinfo-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 19:29:04.082000 [ 635 ] DEBUG : Command:[docker container list --all --filter name='^/roottestmasksensitiveinfo-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 19:29:04.109000 [ 635 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 19:29:04.109000 [ 635 ] DEBUG : No running containers for project: roottestmasksensitiveinfo-gw0 (cluster.py:889, cleanup) 2026-04-10 19:29:04.109000 [ 635 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 19:29:04.138000 [ 635 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 19:29:04.139000 [ 635 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 19:29:04.164000 [ 635 ] DEBUG : Stderr:Error response from daemon: a prune operation is already running (cluster.py:147, run_and_check) 2026-04-10 19:29:04.165000 [ 635 ] DEBUG : Exitcode:1 (cluster.py:149, run_and_check) 2026-04-10 19:29:04.165000 [ 635 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 19:29:04.165000 [ 635 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 19:29:04.195000 [ 635 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 19:29:04.196000 [ 635 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) 2026-04-10 19:29:04.196000 [ 635 ] DEBUG : Setup directory for instance: node (cluster.py:2768, start) 2026-04-10 19:29:04.197000 [ 635 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4638, create_dir) 2026-04-10 19:29:04.197000 [ 635 ] DEBUG : Create directory for common tests configuration (cluster.py:4643, create_dir) 2026-04-10 19:29:04.198000 [ 635 ] DEBUG : Copy common configuration from helpers (cluster.py:4663, create_dir) 2026-04-10 19:29:04.198000 [ 635 ] DEBUG : Generate and write macros file (cluster.py:4715, create_dir) 2026-04-10 19:29:04.199000 [ 635 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_mask_sensitive_info/configs/overrides.xml'] to /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/configs/config.d (cluster.py:4755, create_dir) 2026-04-10 19:29:04.200000 [ 635 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/database (cluster.py:4772, create_dir) 2026-04-10 19:29:04.200000 [ 635 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/logs (cluster.py:4783, create_dir) 2026-04-10 19:29:04.200000 [ 635 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2026-04-10 19:29:04.201000 [ 635 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'LSAN_OPTIONS': 'suppressions=/etc/clickhouse-server/lsan_suppressions.txt', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:5ccda723c1fc', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/coordination', 'AZURITE_PORT': '30000', 'AZURITE_STORAGE_ACCOUNT_URL': 'http://azurite1:30000/devstoreaccount1', 'AZURITE_CONNECTION_STRING': 'DefaultEndpointsProtocol=http;AccountName=devstoreaccount1;AccountKey=Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==;BlobEndpoint=http://azurite1:30000/devstoreaccount1;'} stored in /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env (cluster.py:96, _create_env_file) 2026-04-10 19:29:04.201000 [ 635 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 19:29:04.201000 [ 635 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 19:29:04.202000 [ 635 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2026-04-10 19:29:04.202000 [ 635 ] DEBUG : No config file found (config.py:28, find_config_file) 2026-04-10 19:29:04.215000 [ 635 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2026-04-10 19:29:04.216000 [ 635 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml pull] (cluster.py:121, run_and_check) 2026-04-10 19:30:05.578000 [ 635 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by node (cluster.py:147, run_and_check) 2026-04-10 19:30:05.579000 [ 635 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by node (cluster.py:147, run_and_check) 2026-04-10 19:30:05.579000 [ 635 ] DEBUG : Stderr: zoo3 Skipped - Image is already being pulled by node (cluster.py:147, run_and_check) 2026-04-10 19:30:05.579000 [ 635 ] DEBUG : Stderr: node Pulling (cluster.py:147, run_and_check) 2026-04-10 19:30:05.579000 [ 635 ] DEBUG : Stderr: azurite1 Pulling (cluster.py:147, run_and_check) 2026-04-10 19:30:05.579000 [ 635 ] DEBUG : Stderr: node Pulled (cluster.py:147, run_and_check) 2026-04-10 19:30:05.580000 [ 635 ] DEBUG : Stderr: azurite1 Pulled (cluster.py:147, run_and_check) 2026-04-10 19:30:05.580000 [ 635 ] DEBUG : Setup ZooKeeper (cluster.py:2809, start) 2026-04-10 19:30:05.580000 [ 635 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/log', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/config', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper1/coordination', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/log', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/config', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper2/coordination', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/log', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/config', '/ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/keeper3/coordination'] (cluster.py:2810, start) 2026-04-10 19:30:05.582000 [ 635 ] DEBUG : Command:[docker compose --project-name roottestmasksensitiveinfo-gw0 --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:121, run_and_check) 2026-04-10 19:30:06.448000 [ 635 ] DEBUG : Stderr:time="2026-04-10T19:30:05Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:147, run_and_check) 2026-04-10 19:30:06.448000 [ 635 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Creating (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Created (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:30:06.449000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:30:06.450000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:30:06.450000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:30:06.450000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:30:06.450000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:30:06.450000 [ 635 ] DEBUG : Stderr:time="2026-04-10T19:30:06Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 19:30:06.450000 [ 635 ] DEBUG : Stderr:time="2026-04-10T19:30:06Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 19:30:06.450000 [ 635 ] DEBUG : Wait ZooKeeper to start (cluster.py:2446, wait_zookeeper_to_start) 2026-04-10 19:30:06.451000 [ 635 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2015, get_instance_ip) 2026-04-10 19:30:06.454000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:06.455000 [ 635 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.2.3, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 19:30:06.457000 [ 635 ] INFO : Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:06.458000 [ 635 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:30:06.550000 [ 635 ] INFO : Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:06.551000 [ 635 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:30:06.680000 [ 635 ] INFO : Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:06.681000 [ 635 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:30:06.953000 [ 635 ] INFO : Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:06.954000 [ 635 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:30:07.507000 [ 635 ] INFO : Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:07.508000 [ 635 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:30:08.368000 [ 635 ] INFO : Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:08.369000 [ 635 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2026-04-10 19:30:10.770000 [ 635 ] INFO : Connecting to 172.16.2.3(172.16.2.3):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:10.771000 [ 635 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 19:30:10.779000 [ 635 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 19:30:10.780000 [ 635 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 19:30:10.780000 [ 635 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 19:30:10.781000 [ 635 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 19:30:10.783000 [ 635 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 19:30:10.784000 [ 635 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 19:30:10.784000 [ 635 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 19:30:10.884000 [ 635 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 19:30:10.885000 [ 635 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 19:30:10.885000 [ 635 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2015, get_instance_ip) 2026-04-10 19:30:10.889000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:10.889000 [ 635 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.2.4, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 19:30:10.891000 [ 635 ] INFO : Connecting to 172.16.2.4(172.16.2.4):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:10.892000 [ 635 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 19:30:10.896000 [ 635 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 19:30:10.897000 [ 635 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 19:30:10.898000 [ 635 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 19:30:10.898000 [ 635 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 19:30:10.900000 [ 635 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 19:30:10.901000 [ 635 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 19:30:10.901000 [ 635 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 19:30:11.001000 [ 635 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 19:30:11.002000 [ 635 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 19:30:11.002000 [ 635 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2015, get_instance_ip) 2026-04-10 19:30:11.005000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:11.006000 [ 635 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.2.2, port:2181, use_ssl:False (cluster.py:3322, get_kazoo_client) 2026-04-10 19:30:11.007000 [ 635 ] INFO : Connecting to 172.16.2.2(172.16.2.2):2181, use_ssl: False (connection.py:650, _connect) 2026-04-10 19:30:11.008000 [ 635 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2026-04-10 19:30:11.012000 [ 635 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2026-04-10 19:30:11.013000 [ 635 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2026-04-10 19:30:11.014000 [ 635 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2026-04-10 19:30:11.014000 [ 635 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2026-04-10 19:30:11.016000 [ 635 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2026-04-10 19:30:11.017000 [ 635 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2026-04-10 19:30:11.017000 [ 635 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2026-04-10 19:30:11.117000 [ 635 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2026-04-10 19:30:11.118000 [ 635 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2026-04-10 19:30:11.118000 [ 635 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2462, wait_zookeeper_nodes_to_start) 2026-04-10 19:30:11.118000 [ 635 ] INFO : Trying to create Azurite instance by command docker compose --project-name roottestmasksensitiveinfo-gw0 --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d (cluster.py:3092, start) 2026-04-10 19:30:11.118000 [ 635 ] DEBUG : Command:[docker compose --project-name roottestmasksensitiveinfo-gw0 --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml --verbose up -d] (cluster.py:121, run_and_check) 2026-04-10 19:30:11.659000 [ 635 ] DEBUG : Stderr:time="2026-04-10T19:30:11Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:147, run_and_check) 2026-04-10 19:30:11.659000 [ 635 ] DEBUG : Stderr: Volume "roottestmasksensitiveinfo-gw0_data1-1" Creating (cluster.py:147, run_and_check) 2026-04-10 19:30:11.659000 [ 635 ] DEBUG : Stderr: Volume "roottestmasksensitiveinfo-gw0_data1-1" Created (cluster.py:147, run_and_check) 2026-04-10 19:30:11.659000 [ 635 ] DEBUG : Stderr:time="2026-04-10T19:30:11Z" level=warning msg="Found orphan containers ([roottestmasksensitiveinfo-gw0-zoo3-1 roottestmasksensitiveinfo-gw0-zoo2-1 roottestmasksensitiveinfo-gw0-zoo1-1]) for this project. If you removed or renamed this service in your compose file, you can run this command with the --remove-orphans flag to clean it up." (cluster.py:147, run_and_check) 2026-04-10 19:30:11.659000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:30:11.659000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:30:11.659000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:30:11.660000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:30:11.660000 [ 635 ] DEBUG : Stderr:time="2026-04-10T19:30:11Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 19:30:11.660000 [ 635 ] DEBUG : Stderr:time="2026-04-10T19:30:11Z" level=debug msg="otel error" error="" (cluster.py:147, run_and_check) 2026-04-10 19:30:11.660000 [ 635 ] INFO : Trying to connect to Azurite (cluster.py:3106, start) 2026-04-10 19:30:12.745000 [ 635 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/?restype=REDACTED&comp=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': 'b1b7a36c-3513-11f1-a3f5-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2026-04-10 19:30:12.748000 [ 635 ] DEBUG : Starting new HTTP connection (1): 127.0.0.1:30000 (connectionpool.py:245, _new_conn) 2026-04-10 19:30:12.789000 [ 635 ] DEBUG : http://127.0.0.1:30000 "GET /devstoreaccount1/?restype=account&comp=properties HTTP/1.1" 200 0 (connectionpool.py:547, _make_request) 2026-04-10 19:30:12.790000 [ 635 ] INFO : Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.35.0' 'x-ms-client-request-id': 'b1b7a36c-3513-11f1-a3f5-0242ac110002' 'x-ms-request-id': '32994dde-6ed7-4dfb-ad4a-4d844d228402' 'x-ms-version': 'REDACTED' 'date': 'Fri, 10 Apr 2026 19:30:12 GMT' 'x-ms-sku-name': 'REDACTED' 'x-ms-account-kind': 'REDACTED' 'x-ms-is-hns-enabled': 'REDACTED' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' (_universal.py:550, on_response) 2026-04-10 19:30:12.791000 [ 635 ] DEBUG : {'client_request_id': 'b1b7a36c-3513-11f1-a3f5-0242ac110002', 'request_id': '32994dde-6ed7-4dfb-ad4a-4d844d228402', 'version': '2025-11-05', 'date': datetime.datetime(2026, 4, 10, 19, 30, 12, tzinfo=datetime.timezone.utc), 'sku_name': 'Standard_RAGRS', 'account_kind': 'StorageV2', 'is_hns_enabled': False} (cluster.py:2620, wait_azurite_to_start) 2026-04-10 19:30:12.792000 [ 635 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/?comp=REDACTED&prefix=REDACTED&include=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': 'b1bee6ae-3513-11f1-a3f5-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2026-04-10 19:30:12.801000 [ 635 ] DEBUG : http://127.0.0.1:30000 "GET /devstoreaccount1/?comp=list&prefix=azurite-container&include= HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:12.802000 [ 635 ] INFO : Response status: 200 Response headers: 'Server': 'Azurite-Blob/3.35.0' 'x-ms-client-request-id': 'b1bee6ae-3513-11f1-a3f5-0242ac110002' 'x-ms-request-id': '27897e67-77de-4eb5-88e2-6a4ff1257b41' 'x-ms-version': 'REDACTED' 'content-type': 'application/xml' 'Date': 'Fri, 10 Apr 2026 19:30:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' (_universal.py:550, on_response) 2026-04-10 19:30:12.806000 [ 635 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'GET' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': 'b1c0e846-3513-11f1-a3f5-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2026-04-10 19:30:12.817000 [ 635 ] DEBUG : http://127.0.0.1:30000 "GET /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 404 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:12.817000 [ 635 ] INFO : Response status: 404 Response headers: 'Server': 'Azurite-Blob/3.35.0' 'x-ms-error-code': 'ContainerNotFound' 'x-ms-request-id': '7afa908c-2a0e-4d51-997b-e30fec48e228' 'content-type': 'application/xml' 'Date': 'Fri, 10 Apr 2026 19:30:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Transfer-Encoding': 'chunked' (_universal.py:550, on_response) 2026-04-10 19:30:12.818000 [ 635 ] DEBUG : azurite container 'azurite-container' doesn't exist, creating it (cluster.py:2642, wait_azurite_to_start) 2026-04-10 19:30:12.819000 [ 635 ] INFO : Request URL: 'http://127.0.0.1:30000/devstoreaccount1/azurite-container?restype=REDACTED' Request method: 'PUT' Request headers: 'x-ms-version': 'REDACTED' 'Accept': 'application/xml' 'User-Agent': 'azsdk-python-storage-blob/12.19.0 Python/3.10.12 (Linux-5.15.0-130-generic-x86_64-with-glibc2.35)' 'x-ms-date': 'REDACTED' 'x-ms-client-request-id': 'b1c2f564-3513-11f1-a3f5-0242ac110002' 'Authorization': 'REDACTED' No body was attached to the request (_universal.py:514, on_request) 2026-04-10 19:30:12.829000 [ 635 ] DEBUG : http://127.0.0.1:30000 "PUT /devstoreaccount1/azurite-container?restype=container HTTP/1.1" 201 0 (connectionpool.py:547, _make_request) 2026-04-10 19:30:12.829000 [ 635 ] INFO : Response status: 201 Response headers: 'Server': 'Azurite-Blob/3.35.0' 'etag': '"0x209C7BC32F502C0"' 'last-modified': 'Fri, 10 Apr 2026 19:30:12 GMT' 'x-ms-client-request-id': 'b1c2f564-3513-11f1-a3f5-0242ac110002' 'x-ms-request-id': '0bd9a45a-7ae3-466f-ae1d-bc78194f7071' 'x-ms-version': 'REDACTED' 'Date': 'Fri, 10 Apr 2026 19:30:12 GMT' 'Connection': 'keep-alive' 'Keep-Alive': 'REDACTED' 'Content-Length': '0' (_universal.py:550, on_response) 2026-04-10 19:30:12.830000 [ 635 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml up -d --no-recreate') (cluster.py:3149, start) 2026-04-10 19:30:12.830000 [ 635 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml up -d --no-recreate] (cluster.py:121, run_and_check) 2026-04-10 19:30:13.326000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Running (cluster.py:147, run_and_check) 2026-04-10 19:30:13.326000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Running (cluster.py:147, run_and_check) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Running (cluster.py:147, run_and_check) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Running (cluster.py:147, run_and_check) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Creating (cluster.py:147, run_and_check) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Created (cluster.py:147, run_and_check) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Starting (cluster.py:147, run_and_check) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Started (cluster.py:147, run_and_check) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : ClickHouse instance created (cluster.py:3157, start) 2026-04-10 19:30:13.327000 [ 635 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2015, get_instance_ip) 2026-04-10 19:30:13.330000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.331000 [ 635 ] DEBUG : get_instance_ip instance_name=node (cluster.py:2025, get_instance_global_ipv6) 2026-04-10 19:30:13.332000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.333000 [ 635 ] DEBUG : Waiting for ClickHouse start in node, ip: 172.16.2.6... (cluster.py:3165, start) 2026-04-10 19:30:13.334000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestmasksensitiveinfo-gw0-node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.337000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/8da62ddcdf1b846403a9e4b4eb85920adb04a867fd09eaf37483b8d487ea13f7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.440000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/8da62ddcdf1b846403a9e4b4eb85920adb04a867fd09eaf37483b8d487ea13f7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.545000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/8da62ddcdf1b846403a9e4b4eb85920adb04a867fd09eaf37483b8d487ea13f7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.649000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/8da62ddcdf1b846403a9e4b4eb85920adb04a867fd09eaf37483b8d487ea13f7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.752000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/8da62ddcdf1b846403a9e4b4eb85920adb04a867fd09eaf37483b8d487ea13f7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.856000 [ 635 ] DEBUG : http://localhost:None "GET /v1.46/containers/8da62ddcdf1b846403a9e4b4eb85920adb04a867fd09eaf37483b8d487ea13f7/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2026-04-10 19:30:13.857000 [ 635 ] DEBUG : ClickHouse node started (cluster.py:3169, start) ------------------------------ Captured log call ------------------------------- 2026-04-10 19:30:13.862000 [ 635 ] DEBUG : Executing query CREATE TABLE table0 (x int) ENGINE = MySQL('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', 'EI9HJYOR677QVQBS') on node (cluster.py:3658, query) 2026-04-10 19:30:14.030000 [ 635 ] DEBUG : Executing query CREATE TABLE table1 (x int) ENGINE = PostgreSQL('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', 'EI9HJYOR677QVQBS') on node (cluster.py:3658, query) 2026-04-10 19:30:14.146000 [ 635 ] DEBUG : Executing query CREATE TABLE table2 (x int) ENGINE = MongoDB('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', 'EI9HJYOR677QVQBS') on node (cluster.py:3658, query) 2026-04-10 19:30:14.262000 [ 635 ] DEBUG : Executing query CREATE TABLE table3 (x int) ENGINE = S3('http://minio1:9001/root/data/test1.csv') on node (cluster.py:3658, query) 2026-04-10 19:30:14.379000 [ 635 ] DEBUG : Executing query CREATE TABLE table4 (x int) ENGINE = S3('http://minio1:9001/root/data/test2.csv', 'CSV') on node (cluster.py:3658, query) 2026-04-10 19:30:14.495000 [ 635 ] DEBUG : Executing query CREATE TABLE table5 (x int) ENGINE = S3('http://minio1:9001/root/data/test3.csv.gz', 'CSV', 'gzip') on node (cluster.py:3658, query) 2026-04-10 19:30:14.562000 [ 635 ] DEBUG : Executing query CREATE TABLE table6 (x int) ENGINE = S3('http://minio1:9001/root/data/test4.csv', 'minio', 'EI9HJYOR677QVQBS', 'CSV') on node (cluster.py:3658, query) 2026-04-10 19:30:14.678000 [ 635 ] DEBUG : Executing query CREATE TABLE table7 (x int) ENGINE = S3('http://minio1:9001/root/data/test5.csv.gz', 'minio', 'EI9HJYOR677QVQBS', 'CSV', 'gzip') on node (cluster.py:3658, query) 2026-04-10 19:30:14.794000 [ 635 ] DEBUG : Executing query CREATE TABLE table8 (x int) ENGINE = MySQL(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', table = 'mysql_table', user = 'mysql_user', password = 'EI9HJYOR677QVQBS') on node (cluster.py:3658, query) 2026-04-10 19:30:14.910000 [ 635 ] DEBUG : Executing query CREATE TABLE table9 (x int) ENGINE = MySQL(named_collection_2, database = 'mysql_db', host = 'mysql80', port = 3306, password = 'EI9HJYOR677QVQBS', table = 'mysql_table', user = 'mysql_user') on node (cluster.py:3658, query) 2026-04-10 19:30:14.976000 [ 635 ] DEBUG : Executing query CREATE TABLE table10 (x int) ENGINE = MySQL(named_collection_3, database = 'mysql_db', host = 'mysql80', port = 3306, table = 'mysql_table') on node (cluster.py:3658, query) 2026-04-10 19:30:15.092000 [ 635 ] DEBUG : Executing query CREATE TABLE table11 (x int) ENGINE = PostgreSQL(named_collection_4, host = 'postgres1', port = 5432, database = 'postgres_db', table = 'postgres_table', user = 'postgres_user', password = 'EI9HJYOR677QVQBS') on node (cluster.py:3658, query) 2026-04-10 19:30:15.208000 [ 635 ] DEBUG : Executing query CREATE TABLE table12 (x int) ENGINE = MongoDB(named_collection_5, host = 'mongo1', port = 5432, database = 'mongo_db', collection = 'mongo_col', user = 'mongo_user', password = 'EI9HJYOR677QVQBS') on node (cluster.py:3658, query) 2026-04-10 19:30:15.325000 [ 635 ] DEBUG : Executing query CREATE TABLE table13 (x int) ENGINE = S3(named_collection_6, url = 'http://minio1:9001/root/data/test8.csv', access_key_id = 'minio', secret_access_key = 'EI9HJYOR677QVQBS', format = 'CSV') on node (cluster.py:3658, query) 2026-04-10 19:30:15.442000 [ 635 ] DEBUG : Executing query CREATE TABLE table14 (x int) ENGINE = S3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV', 'gzip') on node (cluster.py:3658, query) 2026-04-10 19:30:15.558000 [ 635 ] DEBUG : Executing query CREATE TABLE table15 (x int) ENGINE = S3('http://minio1:9001/root/data/test10.csv.gz', 'minio', 'EI9HJYOR677QVQBS') on node (cluster.py:3658, query) 2026-04-10 19:30:15.674000 [ 635 ] DEBUG : Executing query CREATE TABLE table16 (x int) ENGINE = DeltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'EI9HJYOR677QVQBS') on node (cluster.py:3741, query_and_get_error) _____________________________ test_table_functions _____________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 def test_table_functions(): password = new_password() azure_conn_string = cluster.env_variables["AZURITE_CONNECTION_STRING"] account_key_pattern = re.compile("AccountKey=.*?(;|$)") masked_azure_conn_string = re.sub( account_key_pattern, "AccountKey=[HIDDEN]\\1", azure_conn_string ) azure_storage_account_url = cluster.env_variables["AZURITE_STORAGE_ACCOUNT_URL"] azure_account_name = "devstoreaccount1" azure_account_key = "Eby8vdM02xNOcqFlqUwJPLlmEtlCDXJ1OUzFT50uSRZ6IFsuFq2UVErCz4I6tq/K1SZFPTOtr/KBHBeksoGMGw==" table_functions = [ f"mysql('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', '{password}')", f"postgresql('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', '{password}')", f"mongodb('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', '{password}', 'x int')", f"s3('http://minio1:9001/root/data/test1.csv')", f"s3('http://minio1:9001/root/data/test2.csv', 'CSV')", f"s3('http://minio1:9001/root/data/test3.csv', 'minio', '{password}')", f"s3('http://minio1:9001/root/data/test4.csv', 'CSV', 'x int')", f"s3('http://minio1:9001/root/data/test5.csv.gz', 'CSV', 'x int', 'gzip')", f"s3('http://minio1:9001/root/data/test6.csv', 'minio', '{password}', 'CSV')", f"s3('http://minio1:9001/root/data/test7.csv', 'minio', '{password}', 'CSV', 'x int')", f"s3('http://minio1:9001/root/data/test8.csv.gz', 'minio', '{password}', 'CSV', 'x int', 'gzip')", f"remote('127.{{2..11}}', default.remote_table)", f"remote('127.{{2..11}}', default.remote_table, rand())", f"remote('127.{{2..11}}', default.remote_table, 'remote_user')", f"remote('127.{{2..11}}', default.remote_table, 'remote_user', '{password}')", f"remote('127.{{2..11}}', default.remote_table, 'remote_user', rand())", f"remote('127.{{2..11}}', default.remote_table, 'remote_user', '{password}', rand())", f"remote('127.{{2..11}}', 'default.remote_table', 'remote_user', '{password}', rand())", f"remote('127.{{2..11}}', 'default', 'remote_table', 'remote_user', '{password}', rand())", f"remote('127.{{2..11}}', numbers(10), 'remote_user', '{password}', rand())", f"remoteSecure('127.{{2..11}}', 'default', 'remote_table', 'remote_user', '{password}')", f"remoteSecure('127.{{2..11}}', 'default', 'remote_table', 'remote_user', rand())", f"mysql(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', table = 'mysql_table', user = 'mysql_user', password = '{password}')", f"postgresql(named_collection_2, password = '{password}', host = 'postgres1', port = 5432, database = 'postgres_db', table = 'postgres_table', user = 'postgres_user')", f"s3(named_collection_2, url = 'http://minio1:9001/root/data/test4.csv', access_key_id = 'minio', secret_access_key = '{password}')", f"remote(named_collection_6, addresses_expr = '127.{{2..11}}', database = 'default', table = 'remote_table', user = 'remote_user', password = '{password}', sharding_key = rand())", f"remoteSecure(named_collection_6, addresses_expr = '127.{{2..11}}', database = 'default', table = 'remote_table', user = 'remote_user', password = '{password}')", f"s3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV')", f"s3('http://minio1:9001/root/data/test10.csv.gz', 'minio', '{password}')", f"deltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", f"azureBlobStorage('{azure_conn_string}', 'cont', 'test_simple.csv', 'CSV')", f"azureBlobStorage('{azure_conn_string}', 'cont', 'test_simple_1.csv', 'CSV', 'none')", f"azureBlobStorage('{azure_conn_string}', 'cont', 'test_simple_2.csv', 'CSV', 'none', 'auto')", f"azureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_3.csv', '{azure_account_name}', '{azure_account_key}')", f"azureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_4.csv', '{azure_account_name}', '{azure_account_key}', 'CSV')", f"azureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_5.csv', '{azure_account_name}', '{azure_account_key}', 'CSV', 'none')", f"azureBlobStorage('{azure_storage_account_url}', 'cont', 'test_simple_6.csv', '{azure_account_name}', '{azure_account_key}', 'CSV', 'none', 'auto')", f"azureBlobStorage(named_collection_2, connection_string = '{azure_conn_string}', container = 'cont', blob_path = 'test_simple_7.csv', format = 'CSV')", f"azureBlobStorage(named_collection_2, storage_account_url = '{azure_storage_account_url}', container = 'cont', blob_path = 'test_simple_8.csv', account_name = '{azure_account_name}', account_key = '{azure_account_key}')", f"iceberg('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", f"gcs('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", f"icebergS3('http://minio1:9001/root/data/test11.csv.gz', 'minio', '{password}')", f"icebergAzure('{azure_storage_account_url}', 'cont', 'test_simple_6.csv', '{azure_account_name}', '{azure_account_key}', 'CSV', 'none', 'auto')", ] def make_test_case(i): table_name = f"tablefunc{i}" table_function = table_functions[i] error = None if isinstance(table_function, tuple): table_function, error = table_function query = f"CREATE TABLE {table_name} (x int) AS {table_function}" return table_name, query, error # Generate test cases as a list of tuples (table_name, query, error). test_cases = [make_test_case(i) for i in range(len(table_functions))] for table_name, query, error in test_cases: > node.query(query) test_mask_sensitive_info/test.py:503: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/cluster.py:3659: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 46, stderr: Received exception from server (version 25.3.8): E Code: 46. DB::Exception: Received from 172.16.2.6:9000. DB::Exception: Unknown table function deltaLake. Stack trace: E E 0. DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000d7ed77b E 1. DB::Exception::Exception(PreformattedMessage&&, int) @ 0x0000000008292ccc E 2. DB::Exception::Exception(int, FormatStringHelperImpl::type>, String const&) @ 0x0000000008b8f68b E 3. DB::TableFunctionFactory::get(std::shared_ptr const&, std::shared_ptr) const @ 0x0000000010068628 E 4. DB::InterpreterCreateQuery::doCreateTable(DB::ASTCreateQuery&, DB::InterpreterCreateQuery::TableProperties const&, std::unique_ptr>&, DB::LoadingStrictnessLevel) @ 0x000000001187a81b E 5. DB::InterpreterCreateQuery::createTable(DB::ASTCreateQuery&) @ 0x000000001186f6c2 E 6. DB::InterpreterCreateQuery::execute() @ 0x0000000011882c18 E 7. DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*, std::shared_ptr&) @ 0x0000000011cadfab E 8. DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000011ca89c4 E 9. DB::TCPHandler::runImpl() @ 0x0000000012fed5ac E 10. DB::TCPHandler::run() @ 0x000000001300b799 E 11. Poco::Net::TCPServerConnection::start() @ 0x0000000016744027 E 12. Poco::Net::TCPServerDispatcher::run() @ 0x0000000016744479 E 13. Poco::PooledThread::run() @ 0x00000000167107db E 14. Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001670ecbd E 15. ? @ 0x00007ff928f97ac3 E 16. ? @ 0x00007ff929029850 E . (UNKNOWN_FUNCTION) E (query: CREATE TABLE tablefunc29 (x int) AS deltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'QZQQDDGNR2PWTKP4')) helpers/client.py:248: QueryRuntimeException ------------------------------ Captured log call ------------------------------- 2026-04-10 19:30:15.906000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc0 (x int) AS mysql('mysql80:3306', 'mysql_db', 'mysql_table', 'mysql_user', 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:16.022000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc1 (x int) AS postgresql('postgres1:5432', 'postgres_db', 'postgres_table', 'postgres_user', 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:16.138000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc2 (x int) AS mongodb('mongo1:27017', 'mongo_db', 'mongo_col', 'mongo_user', 'QZQQDDGNR2PWTKP4', 'x int') on node (cluster.py:3658, query) 2026-04-10 19:30:16.255000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc3 (x int) AS s3('http://minio1:9001/root/data/test1.csv') on node (cluster.py:3658, query) 2026-04-10 19:30:16.371000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc4 (x int) AS s3('http://minio1:9001/root/data/test2.csv', 'CSV') on node (cluster.py:3658, query) 2026-04-10 19:30:16.487000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc5 (x int) AS s3('http://minio1:9001/root/data/test3.csv', 'minio', 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:16.553000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc6 (x int) AS s3('http://minio1:9001/root/data/test4.csv', 'CSV', 'x int') on node (cluster.py:3658, query) 2026-04-10 19:30:16.669000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc7 (x int) AS s3('http://minio1:9001/root/data/test5.csv.gz', 'CSV', 'x int', 'gzip') on node (cluster.py:3658, query) 2026-04-10 19:30:16.785000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc8 (x int) AS s3('http://minio1:9001/root/data/test6.csv', 'minio', 'QZQQDDGNR2PWTKP4', 'CSV') on node (cluster.py:3658, query) 2026-04-10 19:30:16.901000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc9 (x int) AS s3('http://minio1:9001/root/data/test7.csv', 'minio', 'QZQQDDGNR2PWTKP4', 'CSV', 'x int') on node (cluster.py:3658, query) 2026-04-10 19:30:17.017000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc10 (x int) AS s3('http://minio1:9001/root/data/test8.csv.gz', 'minio', 'QZQQDDGNR2PWTKP4', 'CSV', 'x int', 'gzip') on node (cluster.py:3658, query) 2026-04-10 19:30:17.084000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc11 (x int) AS remote('127.{2..11}', default.remote_table) on node (cluster.py:3658, query) 2026-04-10 19:30:17.200000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc12 (x int) AS remote('127.{2..11}', default.remote_table, rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:17.316000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc13 (x int) AS remote('127.{2..11}', default.remote_table, 'remote_user') on node (cluster.py:3658, query) 2026-04-10 19:30:17.432000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc14 (x int) AS remote('127.{2..11}', default.remote_table, 'remote_user', 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:17.549000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc15 (x int) AS remote('127.{2..11}', default.remote_table, 'remote_user', rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:17.666000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc16 (x int) AS remote('127.{2..11}', default.remote_table, 'remote_user', 'QZQQDDGNR2PWTKP4', rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:17.782000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc17 (x int) AS remote('127.{2..11}', 'default.remote_table', 'remote_user', 'QZQQDDGNR2PWTKP4', rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:17.898000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc18 (x int) AS remote('127.{2..11}', 'default', 'remote_table', 'remote_user', 'QZQQDDGNR2PWTKP4', rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:18.014000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc19 (x int) AS remote('127.{2..11}', numbers(10), 'remote_user', 'QZQQDDGNR2PWTKP4', rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:18.132000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc20 (x int) AS remoteSecure('127.{2..11}', 'default', 'remote_table', 'remote_user', 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:18.248000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc21 (x int) AS remoteSecure('127.{2..11}', 'default', 'remote_table', 'remote_user', rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:18.365000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc22 (x int) AS mysql(named_collection_1, host = 'mysql80', port = 3306, database = 'mysql_db', table = 'mysql_table', user = 'mysql_user', password = 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:18.481000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc23 (x int) AS postgresql(named_collection_2, password = 'QZQQDDGNR2PWTKP4', host = 'postgres1', port = 5432, database = 'postgres_db', table = 'postgres_table', user = 'postgres_user') on node (cluster.py:3658, query) 2026-04-10 19:30:18.597000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc24 (x int) AS s3(named_collection_2, url = 'http://minio1:9001/root/data/test4.csv', access_key_id = 'minio', secret_access_key = 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:18.714000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc25 (x int) AS remote(named_collection_6, addresses_expr = '127.{2..11}', database = 'default', table = 'remote_table', user = 'remote_user', password = 'QZQQDDGNR2PWTKP4', sharding_key = rand()) on node (cluster.py:3658, query) 2026-04-10 19:30:18.830000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc26 (x int) AS remoteSecure(named_collection_6, addresses_expr = '127.{2..11}', database = 'default', table = 'remote_table', user = 'remote_user', password = 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:18.946000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc27 (x int) AS s3('http://minio1:9001/root/data/test9.csv.gz', 'NOSIGN', 'CSV') on node (cluster.py:3658, query) 2026-04-10 19:30:19.063000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc28 (x int) AS s3('http://minio1:9001/root/data/test10.csv.gz', 'minio', 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) 2026-04-10 19:30:19.129000 [ 635 ] DEBUG : Executing query CREATE TABLE tablefunc29 (x int) AS deltaLake('http://minio1:9001/root/data/test11.csv.gz', 'minio', 'QZQQDDGNR2PWTKP4') on node (cluster.py:3658, query) ---------------------------- Captured log teardown ----------------------------- 2026-04-10 19:30:19.421000 [ 635 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml stop --timeout 20] (cluster.py:121, run_and_check) 2026-04-10 19:30:33.202000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.203000 [ 635 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 19:30:33.214000 [ 635 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/.env --project-name roottestmasksensitiveinfo-gw0 --file /ClickHouse/tests/integration/test_mask_sensitive_info/_instances-1-gw0/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_azurite.yml down --volumes] (cluster.py:121, run_and_check) 2026-04-10 19:30:33.784000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.784000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.784000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:30:33.784000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-node-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:30:33.784000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:30:33.785000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo3-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-zoo2-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Container roottestmasksensitiveinfo-gw0-azurite1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:30:33.786000 [ 635 ] DEBUG : Stderr: Volume roottestmasksensitiveinfo-gw0_data1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:30:33.787000 [ 635 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Removing (cluster.py:147, run_and_check) 2026-04-10 19:30:33.787000 [ 635 ] DEBUG : Stderr: Volume roottestmasksensitiveinfo-gw0_data1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:30:33.787000 [ 635 ] DEBUG : Stderr: Network roottestmasksensitiveinfo-gw0_default Removed (cluster.py:147, run_and_check) 2026-04-10 19:30:33.787000 [ 635 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 19:30:33.810000 [ 635 ] DEBUG : Docker networks for project roottestmasksensitiveinfo-gw0 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 19:30:33.840000 [ 635 ] DEBUG : Docker containers for project roottestmasksensitiveinfo-gw0 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 19:30:33.868000 [ 635 ] DEBUG : Docker volumes for project roottestmasksensitiveinfo-gw0 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 19:30:33.868000 [ 635 ] DEBUG : Command:[docker container list --all --filter name='^/roottestmasksensitiveinfo-gw0-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 19:30:33.897000 [ 635 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 19:30:33.897000 [ 635 ] DEBUG : No running containers for project: roottestmasksensitiveinfo-gw0 (cluster.py:889, cleanup) 2026-04-10 19:30:33.897000 [ 635 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 19:30:33.926000 [ 635 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 19:30:33.926000 [ 635 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 19:30:33.969000 [ 635 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 19:30:33.969000 [ 635 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 19:30:33.969000 [ 635 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 19:30:33.970000 [ 635 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 19:30:34.000000 [ 635 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 19:30:34.001000 [ 635 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) ___________________________ test_update_one_cluster ____________________________ [gw1] linux -- Python 3.10.12 /usr/bin/python3 started_cluster = def test_update_one_cluster(started_cluster): send_repeated_query("distributed") send_repeated_query("distributed2") > assert get_errors_count("test_cluster") > 0 E AssertionError: assert 0 > 0 E + where 0 = get_errors_count('test_cluster') test_reload_clusters_config/test.py:202: AssertionError ------------------------------ Captured log call ------------------------------- 2026-04-10 19:31:41.675000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:31:53.839000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:32:05.953000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:32:18.066000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:32:30.180000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:32:42.294000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed2 SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:32:54.410000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed2 SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:33:07.326000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed2 SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:33:19.437000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed2 SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:33:31.500000 [ 638 ] DEBUG : Executing query SELECT count() FROM distributed2 SETTINGS receive_timeout=1, handshake_timeout_ms=1 on node (cluster.py:3741, query_and_get_error) 2026-04-10 19:33:43.613000 [ 638 ] DEBUG : Executing query SELECT errors_count FROM system.clusters WHERE cluster='test_cluster' and host_name='node_1' on node (cluster.py:3658, query) ---------------------------- Captured log teardown ----------------------------- 2026-04-10 19:33:43.812000 [ 638 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_reload_clusters_config/_instances-1-gw1/.env --project-name roottestreloadclustersconfig-gw1 --file /ClickHouse/tests/integration/test_reload_clusters_config/_instances-1-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml stop --timeout 20] (cluster.py:121, run_and_check) 2026-04-10 19:33:51.350000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.350000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.350000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.350000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.351000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.351000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.351000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.351000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.351000 [ 638 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_reload_clusters_config/_instances-1-gw1/node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_reload_clusters_config/_instances-1-gw1/node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:121, run_and_check) 2026-04-10 19:33:51.366000 [ 638 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_reload_clusters_config/_instances-1-gw1/.env --project-name roottestreloadclustersconfig-gw1 --file /ClickHouse/tests/integration/test_reload_clusters_config/_instances-1-gw1/node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml down --volumes] (cluster.py:121, run_and_check) 2026-04-10 19:33:51.901000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-node-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.901000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-node-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.901000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-node-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:33:51.901000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-node-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:33:51.901000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo3-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.901000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo1-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo2-1 Stopping (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo2-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo2-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo3-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo3-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo1-1 Stopped (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo1-1 Removing (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo2-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:33:51.902000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo3-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:33:51.903000 [ 638 ] DEBUG : Stderr: Container roottestreloadclustersconfig-gw1-zoo1-1 Removed (cluster.py:147, run_and_check) 2026-04-10 19:33:51.903000 [ 638 ] DEBUG : Stderr: Network roottestreloadclustersconfig-gw1_default Removing (cluster.py:147, run_and_check) 2026-04-10 19:33:51.903000 [ 638 ] DEBUG : Stderr: Network roottestreloadclustersconfig-gw1_default Removed (cluster.py:147, run_and_check) 2026-04-10 19:33:51.903000 [ 638 ] DEBUG : Cleanup called (cluster.py:861, cleanup) 2026-04-10 19:33:51.926000 [ 638 ] DEBUG : Docker networks for project roottestreloadclustersconfig-gw1 are NETWORK ID NAME DRIVER SCOPE (cluster.py:840, print_all_docker_pieces) 2026-04-10 19:33:51.953000 [ 638 ] DEBUG : Docker containers for project roottestreloadclustersconfig-gw1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:848, print_all_docker_pieces) 2026-04-10 19:33:51.982000 [ 638 ] DEBUG : Docker volumes for project roottestreloadclustersconfig-gw1 are DRIVER VOLUME NAME (cluster.py:856, print_all_docker_pieces) 2026-04-10 19:33:51.982000 [ 638 ] DEBUG : Command:[docker container list --all --filter name='^/roottestreloadclustersconfig-gw1-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:121, run_and_check) 2026-04-10 19:33:52.012000 [ 638 ] DEBUG : Unstopped containers: {} (cluster.py:875, cleanup) 2026-04-10 19:33:52.012000 [ 638 ] DEBUG : No running containers for project: roottestreloadclustersconfig-gw1 (cluster.py:889, cleanup) 2026-04-10 19:33:52.012000 [ 638 ] DEBUG : Trying to prune unused networks... (cluster.py:895, cleanup) 2026-04-10 19:33:52.040000 [ 638 ] DEBUG : Trying to prune unused images... (cluster.py:911, cleanup) 2026-04-10 19:33:52.040000 [ 638 ] DEBUG : Command:[docker image prune -f] (cluster.py:121, run_and_check) 2026-04-10 19:33:52.083000 [ 638 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:145, run_and_check) 2026-04-10 19:33:52.084000 [ 638 ] DEBUG : Images pruned (cluster.py:914, cleanup) 2026-04-10 19:33:52.084000 [ 638 ] DEBUG : Trying to prune unused volumes... (cluster.py:920, cleanup) 2026-04-10 19:33:52.084000 [ 638 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:121, run_and_check) 2026-04-10 19:33:52.114000 [ 638 ] DEBUG : Stdout:1 (cluster.py:145, run_and_check) 2026-04-10 19:33:52.114000 [ 638 ] DEBUG : Volumes pruned: 1 (cluster.py:925, cleanup) ----------------- generated report log file: parallel1_1.jsonl ----------------- ============================== slowest durations =============================== 133.65s call test_reload_clusters_config/test.py::test_delete_cluster 122.06s call test_reload_clusters_config/test.py::test_update_one_cluster 70.01s setup test_mask_sensitive_info/test.py::test_create_table 24.17s setup test_reload_clusters_config/test.py::test_delete_cluster 14.58s teardown test_mask_sensitive_info/test.py::test_table_functions 8.30s teardown test_reload_clusters_config/test.py::test_update_one_cluster 3.34s call test_mask_sensitive_info/test.py::test_table_functions 1.93s call test_mask_sensitive_info/test.py::test_create_table 0.00s teardown test_reload_clusters_config/test.py::test_delete_cluster 0.00s setup test_reload_clusters_config/test.py::test_update_one_cluster 0.00s teardown test_mask_sensitive_info/test.py::test_create_table 0.00s setup test_mask_sensitive_info/test.py::test_table_functions =========================== short test summary info ============================ FAILED test_mask_sensitive_info/test.py::test_create_table - assert 'DNS_ERRO... FAILED test_mask_sensitive_info/test.py::test_table_functions - helpers.clien... FAILED test_reload_clusters_config/test.py::test_update_one_cluster - Asserti... PASSED test_reload_clusters_config/test.py::test_delete_cluster =================== 3 failed, 1 passed in 290.68s (0:04:50) ==================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 492, in subprocess.check_call(cmd, shell=True, bufsize=0) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_af1sbo --privileged --dns-search='.' --memory=30709035008 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=5ccda723c1fc -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=d862517635bf -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=1 --color=no --durations=0 --report-log=parallel1_1.jsonl --report-log-exclude-logs-on-passed-tests test_mask_sensitive_info/test.py::test_create_table test_mask_sensitive_info/test.py::test_table_functions test_reload_clusters_config/test.py::test_delete_cluster test_reload_clusters_config/test.py::test_update_one_cluster -vvv " altinityinfra/integration-tests-runner:226bfaf75ac1 ' returned non-zero exit status 1.